Online Deep Learning Hyperparameter Tuning based on Provenance Analysis
نویسندگان
چکیده
Training Deep Learning (DL) models require adjusting a series of hyperparameters. Although there are several tools to automatically choose the best hyperparameter configuration, user is still main actor take final decision. To decide whether training should continue or try different configurations, needs analyze online hyperparameters most adequate dataset, observing metrics such as accuracy and loss values. Provenance naturally represents data derivation relationships (i.e., transformations, parameter values, etc.), which provide important support in this analysis. Most existing provenance solutions define their own proprietary representations DL users choosing makes analysis interoperability difficult. We present Keras-Prov its extension, named Keras-Prov++, provides an analytical dashboard fine-tuning. Different from current mainstream solutions, captures applications using W3C PROV recommendation, allowing for help deciding on changing hyperparameters’ values after performance validation set. experimental evaluation Keras-Prov++ AlexNet real case study, DenseED, that acts surrogate model solving equations. During analysis, identify scenarios suggest reducing number epochs avoid unnecessary executions fine-tuning learning rate improve accuracy.
منابع مشابه
Collaborative hyperparameter tuning
Hyperparameter learning has traditionally been a manual task because of the limited number of trials. Today’s computing infrastructures allow bigger evaluation budgets, thus opening the way for algorithmic approaches. Recently, surrogate-based optimization was successfully applied to hyperparameter learning for deep belief networks and to WEKA classifiers. The methods combined brute force compu...
متن کاملMassively Parallel Hyperparameter Tuning
Modern machine learning models are characterized by large hyperparameter search spaces and prohibitively expensive training costs. For such models, we cannot afford to train candidate models sequentially and wait months before finding a suitable hyperparameter configuration. Hence, we introduce the large-scale regime for parallel hyperparameter tuning, where we need to evaluate orders of magnit...
متن کاملEfficient Transfer Learning Method for Automatic Hyperparameter Tuning
We propose a fast and effective algorithm for automatic hyperparameter tuning that can generalize across datasets. Our method is an instance of sequential model-based optimization (SMBO) that transfers information by constructing a common response surface for all datasets, similar to Bardenet et al. (2013). The time complexity of reconstructing the response surface at every SMBO iteration in ou...
متن کاملOnline Deep Learning: Learning Deep Neural Networks on the Fly
Deep Neural Networks (DNNs) are typically trained by backpropagation in a batch learning setting, which requires the entire training data to be made available prior to the learning task. This is not scalable for many real-world scenarios where new data arrives sequentially in a stream form. We aim to address an open challenge of “Online Deep Learning” (ODL) for learning DNNs on the fly in an on...
متن کاملA Grouping Hotel Recommender System Based on Deep Learning and Sentiment Analysis
Recommender systems are important tools for users to identify their preferred items and for businesses to improve their products and services. In recent years, the use of online services for selection and reservation of hotels have witnessed a booming growth. Customer’ reviews have replaced the word of mouth marketing, but searching hotels based on user priorities is more time-consuming. This s...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Journal of Information and Data Management
سال: 2021
ISSN: ['2178-7107']
DOI: https://doi.org/10.5753/jidm.2021.1924